The Weighted Sparsity Problem: Complexity and Algorithms
نویسندگان
چکیده
Many optimization algorithms involve repeated processing of a fixed set of linear constraints. If we pre-process the constraint matrix A to make it sparser, algebraic operations should become faster. In many applications there is a priori information about the likelihood that each column will appear in a basis, which can be expressed as weights on the columns. This leads to considering the Weighted Sparsity Problem (WSP): Find a row-equivalent constraint matrix with as small a weight of non-zeros as possible. We show that WSP is NP-hard even with a non-degeneracy assumption, and even if restricted to instances with at most three non-zeros per both row and column. WSP is shown to have a polynomial algorithm when the number of non-zeros per either row or column is limited to at most two. This contrasts with previous results that, assuming only non-degeneracy, the unweighted version of WSP does have a polynomial algorithm (this has proven to be practically useful in tests on real data). The polynomial algorithm for WSP with at most two non-zeros per row or column is based on solving one-row problems via minimum cut calculations, together with a sufficient condition for piecing these one-row solutions together into a global solution.
منابع مشابه
A NOVEL FUZZY-BASED SIMILARITY MEASURE FOR COLLABORATIVE FILTERING TO ALLEVIATE THE SPARSITY PROBLEM
Memory-based collaborative filtering is the most popular approach to build recommender systems. Despite its success in many applications, it still suffers from several major limitations, including data sparsity. Sparse data affect the quality of the user similarity measurement and consequently the quality of the recommender system. In this paper, we propose a novel user similarity measure based...
متن کاملSearch Based Weighted Multi-Bit Flipping Algorithm for High-Performance Low-Complexity Decoding of LDPC Codes
In this paper, two new hybrid algorithms are proposed for decoding Low Density Parity Check (LDPC) codes. Original version of the proposed algorithms named Search Based Weighted Multi Bit Flipping (SWMBF). The main idea of these algorithms is flipping variable multi bits in each iteration, change in which leads to the syndrome vector with least hamming weight. To achieve this, the proposed algo...
متن کاملSearch Based Weighted Multi-Bit Flipping Algorithm for High-Performance Low-Complexity Decoding of LDPC Codes
In this paper, two new hybrid algorithms are proposed for decoding Low Density Parity Check (LDPC) codes. Original version of the proposed algorithms named Search Based Weighted Multi Bit Flipping (SWMBF). The main idea of these algorithms is flipping variable multi bits in each iteration, change in which leads to the syndrome vector with least hamming weight. To achieve this, the proposed algo...
متن کاملApplications of two new algorithms of cuckoo optimization (CO) and forest optimization (FO) for solving single row facility layout problem (SRFLP)
Nowadays, due to inherent complexity of real optimization problems, it has always been a challenging issue to develop a solution algorithm to these problems. Single row facility layout problem (SRFLP) is a NP-hard problem of arranging a number of rectangular facilities with varying length on one side of a straight line with aim of minimizing the weighted sum of the distance between all facility...
متن کاملحل مسئله زمان بندی ماشینهای موازی نامرتبط با اهداف کل زودکرد وزنی و کل دیرکرد وزنی با استفاده از الگوریتم جستجوی پراکنده چند هدفه
The parallel machine scheduling problem is an important and difficult problem to be considered in the real-world situations. Traditionally, this problem consists of the scheduling of a set of independent jobs on parallel machines with the aim of minimizing the maximum job completion. In today's manufacturing systems, in which both early and tardy finishing of job processing are undesired, the o...
متن کاملBlock-Based Compressive Sensing Using Soft Thresholding of Adaptive Transform Coefficients
Compressive sampling (CS) is a new technique for simultaneous sampling and compression of signals in which the sampling rate can be very small under certain conditions. Due to the limited number of samples, image reconstruction based on CS samples is a challenging task. Most of the existing CS image reconstruction methods have a high computational complexity as they are applied on the entire im...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- SIAM J. Discrete Math.
دوره 6 شماره
صفحات -
تاریخ انتشار 1993